Overtraining in neural networks that interpret clinical data.

نویسندگان

  • M L Astion
  • M H Wener
  • R G Thomas
  • G G Hunder
  • D A Bloch
چکیده

Backpropagation neural networks are a computer-based pattern-recognition method that has been applied to the interpretation of clinical data. Unlike rule-based pattern recognition, backpropagation networks learn by being repetitively trained with examples of the patterns to be differentiated. We describe and analyze the phenomenon of overtraining in backpropagation networks. Overtraining refers to the reduction in generalization ability that can occur as networks are trained. The clinical application we used was the differentiation of giant cell arteritis (GCA) from other forms of vasculitis (OTH) based on results for 807 patients (593 OTH, 214 GCA) and eight clinical predictor variables. The 807 cases were randomly assigned to either a training set with 404 cases or to a cross-validation set with the remaining 403 cases. The cross-validation set was used to monitor generalization during training. Results were obtained for eight networks, each derived from a different random assignment of the 807 cases. Training error monotonically decreased during training. In contrast, the cross-validation error usually reached a minimum early in training while the training error was still decreasing. Training beyond the minimum cross-validation error was associated with an increased cross-validation error. The shape of the cross-validation error curve and the point during training corresponding to the minimum cross-validation error varied with the composition of the data sets and the training conditions. The study indicates that training error is not a reliable indicator of a network's ability to generalize. To find the point during training when a network generalizes best, one must monitor cross-validation error separately.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Dynamics of Batch Learning in Multilayer Networks { Overrealizability and Overtraining {

This paper investigates the dynamics of batch learning of multilayer neural networks in the asymptotic case where the number of training data is much larger than the number of parameters. We consider regression problems assuming noisy output data. First, we present experimental results on the behavior in the steepest descent learning of multilayer per-ceptrons and three-layer linear neural netw...

متن کامل

Artificial neural networks: applications in predicting pancreatitis survival

Artificial neural networks are intelligent systems that have successfully been used for prediction in different medical fields. In this study, the efficiency of a neural network for predicting the survival of patients with acute pancreatitis is compared with days-of-survival obtained from patients. A three- layer back-propagation neural network was developed for this purpose. Clinical data (e.g...

متن کامل

Artificial neural networks: applications in predicting pancreatitis survival

Artificial neural networks are intelligent systems that have successfully been used for prediction in different medical fields. In this study, the efficiency of a neural network for predicting the survival of patients with acute pancreatitis is compared with days-of-survival obtained from patients. A three- layer back-propagation neural network was developed for this purpose. Clinical data (e.g...

متن کامل

Learning in Neural Networks and an Integrable System

This paper investigates the dynamics of batch learning of multilayer neural networks in the asymptotic case where the number of training data is much larger than the number of parameters. First, we present experimental results on the behavior in the steepest descent learning of multilayer perceptrons and three-layer linear neural networks. We see in these results that strong overtraining, which...

متن کامل

Dynamics of Batch Learning in Multilayer Neural Networks

We discuss the dynamics of batch learning of multilayer neural networks in the asymptotic limit, where the number of trining data is much larger than the number of parameters, emphasizing on the parameterization redundancy in overrealizable cases. In addition to showing experimental results on overtraining in multilayer perceptrons and three-layer linear neural networks, we theoretically prove ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Clinical chemistry

دوره 39 9  شماره 

صفحات  -

تاریخ انتشار 1993